首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   209篇
  免费   30篇
  2021年   2篇
  2019年   8篇
  2018年   4篇
  2017年   5篇
  2016年   13篇
  2015年   14篇
  2014年   9篇
  2013年   59篇
  2012年   9篇
  2011年   7篇
  2010年   5篇
  2009年   7篇
  2008年   9篇
  2007年   12篇
  2006年   8篇
  2005年   12篇
  2004年   10篇
  2003年   13篇
  2002年   10篇
  2001年   6篇
  2000年   5篇
  1999年   7篇
  1995年   1篇
  1993年   1篇
  1985年   2篇
  1981年   1篇
排序方式: 共有239条查询结果,搜索用时 281 毫秒
21.
Mean residual life is a useful dynamic characteristic to study reliability of a system. It has been widely considered in the literature not only for single unit systems but also for coherent systems. This article is concerned with the study of mean residual life for a coherent system that consists of multiple types of dependent components. In particular, the survival signature based generalized mixture representation is obtained for the survival function of a coherent system and it is used to evaluate the mean residual life function. Furthermore, two mean residual life functions under different conditional events on components’ lifetimes are also defined and studied.  相似文献   
22.
Today, many products are designed and manufactured to function for a long period of time before they fail. Determining product reliability is a great challenge to manufacturers of highly reliable products with only a relatively short period of time available for internal life testing. In particular, it may be difficult to determine optimal burn‐in parameters and characterize the residual life distribution. A promising alternative is to use data on a quality characteristic (QC) whose degradation over time can be related to product failure. Typically, product failure corresponds to the first passage time of the degradation path beyond a critical value. If degradation paths can be modeled properly, one can predict failure time and determine the life distribution without actually observing failures. In this paper, we first use a Wiener process to describe the continuous degradation path of the quality characteristic of the product. A Wiener process allows nonconstant variance and nonzero correlation among data collected at different time points. We propose a decision rule for classifying a unit as normal or weak, and give an economic model for determining the optimal termination time and other parameters of a burn‐in test. Next, we propose a method for assessing the product's lifetime distribution of the passed units. The proposed methodologies are all based only on the product's initial observed degradation data. Finally, an example of an electronic product, namely contact image scanner (CIS), is used to illustrate the proposed procedure. © 2002 Wiley Periodicals, Inc. Naval Research Logistics, 2003  相似文献   
23.
In this article, the Building Evacuation Problem with Shared Information (BEPSI) is formulated as a mixed integer linear program, where the objective is to determine the set of routes along which to send evacuees (supply) from multiple locations throughout a building (sources) to the exits (sinks) such that the total time until all evacuees reach the exits is minimized. The formulation explicitly incorporates the constraints of shared information in providing online instructions to evacuees, ensuring that evacuees departing from an intermediate or source location at a mutual point in time receive common instructions. Arc travel time and capacity, as well as supply at the nodes, are permitted to vary with time and capacity is assumed to be recaptured over time. The BEPSI is shown to be NP‐hard. An exact technique based on Benders decomposition is proposed for its solution. Computational results from numerical experiments on a real‐world network representing a four‐story building are given. Results of experiments employing Benders cuts generated in solving a given problem instance as initial cuts in addressing an updated problem instance are also provided. © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   
24.
25.
Instead of measuring a Wiener degradation or performance process at predetermined time points to track degradation or performance of a product for estimating its lifetime, we propose to obtain the first‐passage times of the process over certain nonfailure thresholds. Based on only these intermediate data, we obtain the uniformly minimum variance unbiased estimator and uniformly most accurate confidence interval for the mean lifetime. For estimating the lifetime distribution function, we propose a modified maximum likelihood estimator and a new estimator and prove that, by increasing the sample size of the intermediate data, these estimators and the above‐mentioned estimator of the mean lifetime can achieve the same levels of accuracy as the estimators assuming one has failure times. Thus, our method of using only intermediate data is useful for highly reliable products when their failure times are difficult to obtain. Furthermore, we show that the proposed new estimator of the lifetime distribution function is more accurate than the standard and modified maximum likelihood estimators. We also obtain approximate confidence intervals for the lifetime distribution function and its percentiles. Finally, we use light‐emitting diodes as an example to illustrate our method and demonstrate how to validate the Wiener assumption during the testing. © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   
26.
Burn‐in is a technique to enhance reliability by eliminating weak items from a population of items having heterogeneous lifetimes. System burn‐in can improve system reliability, but the conditions for system burn‐in to be performed after component burn‐in remain a little understood mathematical challenge. To derive such conditions, we first introduce a general model of heterogeneous system lifetimes, in which the component burn‐in information and assembly problems are related to the prediction of system burn‐in. Many existing system burn‐in models become special cases and two important results are identified. First, heterogeneous system lifetimes can be understood naturally as a consequence of heterogeneous component lifetimes and heterogeneous assembly quality. Second, system burn‐in is effective if assembly quality variation in the components and connections which are arranged in series is greater than a threshold, where the threshold depends on the system structure and component failure rates. © 2003 Wiley Periodicals, Inc. Naval Research Logistics 50: 364–380, 2003.  相似文献   
27.
For computing an optimal (Q, R) or kindred inventory policy, the current literature provides mixed signals on whether or when it is safe to approximate a nonnormal lead‐time‐demand (“LTD”) distribution by a normal distribution. The first part of this paper examines this literature critically to justify why the issue warrants further investigations, while the second part presents reliable evidence showing that the system‐cost penalty for using the normal approximation can be quite serious even when the LTD‐distribution's coefficient of variation is quite low—contrary to the prevalent view of the literature. We also identify situations that will most likely lead to large system‐cost penalty. Our results indicate that, given today's technology, it is worthwhile to estimate an LTD‐distribution's shape more accurately and to compute optimal inventory policies using statistical distributions that more accurately reflect the LTD‐distributions' actual shapes. © 2003 Wiley Periodicals, Inc. Naval Research Logistics, 2003  相似文献   
28.
The coordination of production, supply, and distribution is an important issue in logistics and operations management. This paper develops and analyzes a single‐machine scheduling model that incorporates the scheduling of jobs and the pickup and delivery arrangements of the materials and finished jobs. In this model, there is a capacitated pickup and delivery vehicle that travels between the machine and the storage area, and the objective is to minimize the makespan of the schedule. The problem is strongly NP‐hard in general but is solvable in polynomial time when the job processing sequence is predetermined. An efficient heuristic is developed for the general problem. The effectiveness of the heuristic is studied both analytically and computationally. © 2005 Wiley Periodicals, Inc. Naval Research Logistics, 2005.  相似文献   
29.
An area defense consists of several groups that act independently, i.e., do not communicate with each other. Each group has a fixed number of defenders and a controller that allocates these defenders optimally against the individual attackers comprising an attack. We analyze the effectiveness of this partially coordinated defense against a simultaneous attack of known size in which all attackers are considered to be equally lethal. © 2004 Wiley Periodicals, Inc. Naval Research Logistics, 2005.  相似文献   
30.
We consider the problem of finding the Kth shortest path for a time‐schedule network, where each node in the network has a list of prespecified departure times, and departure from the node can take place only at one of these departure times. We develop a polynomial time algorithm independent of K for finding the Kth shortest path. The proposed algorithm constructs a map structure at each node in the network, using which we can directly find the Kth shortest path without having to enumerate the first K − 1 paths. Since the same map structure is used for different K values, it is not necessary to reconstruct the table for additional paths. Consequently, the algorithm is suitable for directly finding multiple shortest paths in the same network. Furthermore, the algorithm is modified slightly for enumerating the first K shortest paths and is shown to have the lowest possible time complexity under a condition that holds for most practical networks. © 2004 Wiley Periodicals, Inc. Naval Research Logistics, 2005.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号